Recurrence of optimum for training weight and activation quantized networks

نویسندگان

چکیده

Deep neural networks (DNNs) are quantized for efficient inference on resource-constrained platforms. However, training deep learning models with low-precision weights and activations involves a demanding optimization task, which calls minimizing stage-wise loss function subject to discrete set-constraint. While numerous methods have been proposed, existing studies full quantization of DNNs mostly empirical. From theoretical point view, we study practical techniques overcoming the combinatorial nature network quantization. Specifically, investigate simple yet powerful projected gradient-like algorithm quantizing two-layer convolutional networks, by repeatedly moving one step at float in negative direction heuristic fake gradient (so-called coarse gradient) evaluated weights. For first time, prove that under mild conditions, sequence recurrently visit global optimum minimization problem fully network. We also show numerical evidence recurrence phenomenon weight evolution networks.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Role of Heuristic Methods with variable Lengths In ANFIS Networks Optimum Design and Training

ANFIS systems have been much considered due to their acceptable performance in terms of creation of fuzzy classifier and training. One main challenge in designing an ANFIS system is to achieve an efficient method with high accuracy and appropriate interpreting capability. Undoubtedly, type and location of membership functions and the way an ANFIS network is trained are of considerable effect on...

متن کامل

effect of listening strategy training on iranian efl learners listening comprehension and use of strategies

this study aimed to investigate the effects of listening strategy training on iranian efl learners listening comprehension and use of such strategies. this work, employing an experimental methodology, was conducted among 60 adult efl learners from a language institute in isfahan, iran, as participants. the participants, who were selected based on the results of a placement test, were assigned t...

15 صفحه اول

Towards a Deeper Understanding of Training Quantized Neural Networks

Training neural networks with coarsely quantized weights is a key step towards learning on embedded platforms that have limited computing resources, memory capacity, and power consumption. Numerous recent publications have studied methods for training quantized networks, but these studies have been purely experimental. In this work, we investigate the theory of training quantized neural network...

متن کامل

D-Optimum Scanning Node Activation for Large-Scale Monitoring Networks

A method is developed to solve an optimal node activation problem in sensor networks whose measurements are supposed to be used to estimate unknown parameters of the underlying process model. Given a partition of the observation horizon into a finite number of consecutive intervals, the problem is set up to select gaged sites on each interval so that the determinant of the Fisher information ma...

متن کامل

patterns and variations in native and non-native interlanguage pragmatic rating: effects of rater training, intercultural proficiency, and self-assessment

although there are studies on pragmatic assessment, to date, literature has been almost silent about native and non-native english raters’ criteria for the assessment of efl learners’ pragmatic performance. focusing on this topic, this study pursued four purposes. the first one was to find criteria for rating the speech acts of apology and refusal in l2 by native and non-native english teachers...

15 صفحه اول

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Applied and Computational Harmonic Analysis

سال: 2023

ISSN: ['1096-603X', '1063-5203']

DOI: https://doi.org/10.1016/j.acha.2022.07.006